156

Applications in Computer Vision

Algorithm 11 BiRe-ID Training

Input: The training dataset, and the hyper-parameters such as initial learning rate, weight

decay, convolution stride and padding size.

Output: BiRe-ID model with weights bw, learnable scale factors α, and other parameters

p.

1: Initialize w, α, p, and WD randomly;

2: repeat

3:

Randomly sample a mini-batch from dataset;

4:

// Forward propagation

5:

for all i = 1 to N convolution layer do

6:

bai = sign(Φ(αibai1 bwi));

7:

end for

8:

// Backward propagation

9:

for all l = L to 1 do

10:

Update the kernel refining discriminators D(·) of GAN by ascending their stochastic

gradients:

11:

D(log(D(wi; WD)) + log(1D(bwi αi; WD)));

12:

Update the feature refining discriminators D(·) of GAN by ascending their stochas-

tic gradients:

13:

D(log(D(a

H; WD)) + log(1D(aL; WD)));

14:

Calculate the gradients δwi; // Using Eq. 7-12

15:

wiwiη1δwi; // Update the weights

16:

Calculate the gradient δαi; // Using Eq. 13-16

17:

αiαiη2δαi; // Update the scale factor

18:

Calculate the gradient δpi; // Using Eq. 13-16

19:

pipiη3δpi; // Update other parameters

20:

end for

21: until the maximum epoch

22: bw = sign(w).

6.2.5

Ablation Study

In this section, we conduct a performance study for the components of BiRe-ID, including

kernel MSE loss (hyperparameter λ), KR-GAL, feature MSE loss (hyperparameter μ) and

FR-GAL. Market-1501 [289] and ResNet-18 are used in this experiment. We separate this

subsection into two parts: selecting hyperparameters and evaluating the components of

BiRe-ID.

Selecting Hyper-Parameters We first set the kernel refining GAL (KR-GAL) and the

feature refining GAL (FR-GAL) as the invariant variable to compare the impact of the

hyperparameter λ and μ on the ResNet-18 backbone. As plotted in Fig. 6.2, we set the

ablation study at λ and μ. We vary λ from 0 to 1e4 and μ from 0 to 1e2 to evaluate BiRe-

ID’s mAP with different hyperparameter settings. From bottom to top, BiRe-ID obtains

the obviously better mAPs with μ set as 5e3 (green mAP curve). From left to right,

BiRe-ID obtains the best mAP with λ set as 5e5. Therefore, we set μ and λ as 5e3

and 5e5 experiments on the Re-ID task.

Evaluating the Components of BiRe-ID As shown in Table 6.5, the use of GANs

dramatically increases the performance of the proposed baseline network. More specifically,

we first introduce our baseline network by adding a single BN layer ahead of the 1-bit

convolutions of XNOR-Net, which brings a 14.1% improvement in mAP. The introduction

of KR-GAL and FR-GAL improves mAP by 7.1% and 4.1%, respectively, on the proposed